0 bookmark(s) - Sort by: Date ↓ / Title /
Introducing Aeneas, the first AI model for contextualizing ancient inscriptions, designed to help historians better interpret, attribute, and restore fragmentary texts. It reasons across thousands of Latin inscriptions, retrieving textual and contextual parallels to aid in historical research.
The article discusses how agentic LLMs can help users overcome the learning curve of the command line interface (CLI) by automating tasks and providing guidance. It explores tools like ShellGPT and Auto-GPT that leverage LLMs to interpret natural language instructions and execute corresponding CLI commands. The author argues that this approach can make the CLI more accessible and powerful, even for those unfamiliar with its intricacies.
This paper introduces Arch-Router, a preference-aligned routing framework for large language models (LLMs). It addresses limitations in existing routing approaches by focusing on matching queries to user-defined preferences (domain and action types) rather than solely relying on benchmark performance. The framework includes a 1.5B parameter model, Arch-Router, and a data creation pipeline. Experiments demonstrate state-of-the-art results in matching queries with human preferences and improved adaptability.
This article details a step-by-step guide on building a knowledge graph from plain text using an LLM-powered pipeline. It covers concepts like Subject-Predicate-Object triples, text chunking, and LLM prompting to extract structured information.
This article provides a beginner-friendly explanation of attention mechanisms and transformer models, covering sequence-to-sequence modeling, the limitations of RNNs, the concept of attention, and how transformers address these limitations with self-attention and parallelization.
This article provides a comprehensive guide on the basics of BERT (Bidirectional Encoder Representations from Transformers) models. It covers the architecture, use cases, and practical implementations, helping readers understand how to leverage BERT for natural language processing tasks.
The article provides a comprehensive introduction to large language models (LLMs), explaining their purpose, how they function, and their applications. It covers various types of LLMs, including general-purpose and task-specific models, and discusses the distinction between closed-source and open-source LLMs. The article also explores the ethical considerations of building and using LLMs and the future possibilities for these models.
This paper presents a detailed vocabulary of 33 terms and a taxonomy of 58 LLM prompting techniques, along with guidelines for prompt engineering and a meta-analysis of natural language prefix-prompting, serving as the most comprehensive survey on prompt engineering to date.
A tutorial on using LLM for text classification, addressing common challenges and providing practical tips to improve accuracy and usability.
An article discussing the use of embeddings in natural language processing, focusing on comparing open source and closed source embedding models for semantic search, including techniques like clustering and re-ranking.
First / Previous / Next / Last
/ Page 1 of 0